Abuzar Akhtar
B.tech CSE

continuous-probability-distribution-3ed0bc3b-dd22-4699-9264-e20ebd045025

C o n t i n u o u s P r o b a b i l i t y D i s t r i b u t i o n s C o n t i n u o u s P r o b a b i l i t y D i s t r i b u t i o n s ContinuousProbabilityDistributionsContinuous\,Probability\,DistributionsContinuousProbabilityDistributions


A continuous probability distribution function, also known as a probability density function (PDF), describes the probability distribution of a continuous random variable x x xxx. Unlike discrete random variables, which can only take on specific isolated values, continuous random variables can take on any value within a certain range.
The properties of a continuous probability distribution function are as follows:
  1. Non-Negativity: The probability density function is non-negative for all values of x x xxx:
f ( x ) 0 for all x f ( x ) 0 for all  x f(x) >= 0quad"for all "xf(x) \geq 0 \quad \text{for all } xf(x)0for all x
  1. Total Area Under the Curve: The total area under the PDF curve over the entire range is equal to 1, representing the probability of the random variable falling within that range:
f ( x ) d x = 1 f ( x ) d x = 1 int_(-oo)^(oo)f(x)dx=1\int_{-\infty}^{\infty} f(x) \, dx = 1f(x)dx=1
  1. Probability of a Range: The probability that the random variable x x xxx falls within a specific range a x b a x b a <= x <= ba \leq x \leq baxb is given by the integral of the PDF over that range:
P ( a x b ) = a b f ( x ) d x P ( a x b ) = a b f ( x ) d x P(a <= x <= b)=int_(a)^(b)f(x)dxP(a \leq x \leq b) = \int_{a}^{b} f(x) \, dxP(axb)=abf(x)dx
The cumulative distribution function (CDF) is another important concept related to the PDF. It gives the probability that the random variable x x xxx is less than or equal to a given value x 0 x 0 x_(0)x_0x0:
F ( x 0 ) = P ( X x 0 ) = x 0 f ( x ) d x F ( x 0 ) = P ( X x 0 ) = x 0 f ( x ) d x F(x_(0))=P(X <= x_(0))=int_(-oo)^(x_(0))f(x)dxF(x_0) = P(X \leq x_0) = \int_{-\infty}^{x_0} f(x) \, dxF(x0)=P(Xx0)=x0f(x)dx
Examples of Continuous Probability Distribution Functions:
  1. Normal Distribution (Gaussian Distribution).
  2. Uniform Distribution.
  3. Exponential Distribution.
These are just a few examples, and there are many other continuous probability distribution functions used to model various phenomena in different fields of study. Understanding these distributions and their properties is essential for statistical analysis and probability modelling.
=>\Rightarrow Visualizing the Probability Density Function of a Random Variable
The probability density function of a random variable X is given by:
\,

Example:

Suppose you have a continuous random variable x x xxx representing the time (in hours) that a student spends studying for a statistics exam. The probability density function (PDF) of this variable is given by:
f ( x ) = 1 3 e x 3 for x 0 f ( x ) = 1 3 e x 3 for  x 0 f(x)=(1)/(3)e^(-(x)/(3))quad"for "x >= 0f(x) = \frac{1}{3}e^{-\frac{x}{3}} \quad \text{for } x \geq 0f(x)=13ex3for x0
a) Verify that the given function is a valid probability density function (PDF).
b) Find the probability that a student spends between 2 to 5 hours studying for the exam.
c) Calculate the cumulative distribution function (CDF) for x x xxx.
d) What is the probability that a student spends less than 4 hours studying for the exam?
e) If a student is chosen at random, what is the probability that they spend more than 6 hours studying for the exam?
 Solution
a) To verify that the given function is a valid probability density function, we need to check the following properties:
  • Non-negativity: f ( x ) 0 f ( x ) 0 f(x) >= 0f(x) \geq 0f(x)0 for all x 0 x 0 x >= 0x \geq 0x0.
  • Total area under the curve: 0 f ( x ) d x = 1 0 f ( x ) d x = 1 int_(0)^(oo)f(x)dx=1\int_{0}^{\infty} f(x) \, dx = 10f(x)dx=1.
The function f ( x ) = 1 3 e x 3 f ( x ) = 1 3 e x 3 f(x)=(1)/(3)e^(-(x)/(3))f(x) = \frac{1}{3}e^{-\frac{x}{3}}f(x)=13ex3 is non-negative for all x 0 x 0 x >= 0x \geq 0x0. Now, let's calculate the total area under the curve:
0 1 3 e x 3 d x 0 1 3 e x 3 d x int_(0)^(oo)(1)/(3)e^(-(x)/(3))dx\int_{0}^{\infty} \frac{1}{3}e^{-\frac{x}{3}} \, dx013ex3dx
To evaluate this integral, we can use the fact that e a t d t = 1 a e a t + C e a t d t = 1 a e a t + C inte^(-at)dt=-(1)/(a)e^(-at)+C\int e^{-at} \, dt = -\frac{1}{a}e^{-at} + Ceatdt=1aeat+C.
So,
0 1 3 e x 3 d x = [ 1 3 e x 3 ] 0 0 1 3 e x 3 d x = 1 3 e x 3 0 int_(0)^(oo)(1)/(3)e^(-(x)/(3))dx=[-(1)/(3)e^(-(x)/(3))]_(0)^(oo)\int_{0}^{\infty} \frac{1}{3}e^{-\frac{x}{3}} \, dx = \left[-\frac{1}{3}e^{-\frac{x}{3}} \right]_{0}^{\infty}013ex3dx=[13ex3]0
Evaluating at the limits:
[ 1 3 e x 3 ] 0 = 1 3 e 3 ( 1 3 e 0 3 ) 1 3 e x 3 0 = 1 3 e 3 1 3 e 0 3 [-(1)/(3)e^(-(x)/(3))]_(0)^(oo)=-(1)/(3)e^(-(oo)/(3))-(-(1)/(3)e^(-(0)/(3)))\left[-\frac{1}{3}e^{-\frac{x}{3}} \right]_{0}^{\infty} = -\frac{1}{3}e^{-\frac{\infty}{3}} - \left(-\frac{1}{3}e^{-\frac{0}{3}}\right)[13ex3]0=13e3(13e03)
Since e e e^(-oo)e^{-\infty}e approaches zero, we have:
[ 1 3 e x 3 ] 0 = 1 3 1 3 e x 3 0 = 1 3 [-(1)/(3)e^(-(x)/(3))]_(0)^(oo)=(1)/(3)\left[-\frac{1}{3}e^{-\frac{x}{3}} \right]_{0}^{\infty} = \frac{1}{3}[13ex3]0=13
Thus, the total area under the curve is 1, which confirms that the given function is a valid probability density function.
b) To find the probability that a student spends between 2 to 5 hours studying for the exam, we need to calculate:
P ( 2 x 5 ) = 2 5 1 3 e x 3 d x P ( 2 x 5 ) = 2 5 1 3 e x 3 d x P(2 <= x <= 5)=int_(2)^(5)(1)/(3)e^(-(x)/(3))dxP(2 \leq x \leq 5) = \int_{2}^{5} \frac{1}{3}e^{-\frac{x}{3}} \, dxP(2x5)=2513ex3dx
Using the same integration technique as before:
P ( 2 x 5 ) = [ 1 3 e x 3 ] 2 5 = 1 3 e 5 3 + 1 3 e 2 3 P ( 2 x 5 ) = 1 3 e x 3 2 5 = 1 3 e 5 3 + 1 3 e 2 3 P(2 <= x <= 5)=[-(1)/(3)e^(-(x)/(3))]_(2)^(5)=-(1)/(3)e^(-(5)/(3))+(1)/(3)e^(-(2)/(3))P(2 \leq x \leq 5) = \left[-\frac{1}{3}e^{-\frac{x}{3}} \right]_{2}^{5} = -\frac{1}{3}e^{-\frac{5}{3}} + \frac{1}{3}e^{-\frac{2}{3}}P(2x5)=[13ex3]25=13e53+13e23
Calculating the numerical value:
P ( 2 x 5 ) 0.184 P ( 2 x 5 ) 0.184 P(2 <= x <= 5)~~0.184P(2 \leq x \leq 5) \approx 0.184P(2x5)0.184
So, the probability that a student spends between 2 to 5 hours studying for the exam is approximately 0.184.
c) The cumulative distribution function (CDF) for x x xxx is given by:
F ( x ) = x f ( t ) d t F ( x ) = x f ( t ) d t F(x)=int_(-oo)^(x)f(t)dtF(x) = \int_{-\infty}^{x} f(t) \, dtF(x)=xf(t)dt
For our PDF f ( x ) = 1 3 e x 3 f ( x ) = 1 3 e x 3 f(x)=(1)/(3)e^(-(x)/(3))f(x) = \frac{1}{3}e^{-\frac{x}{3}}f(x)=13ex3, the CDF is:
F ( x ) = 0 x 1 3 e t 3 d t F ( x ) = 0 x 1 3 e t 3 d t F(x)=int_(0)^(x)(1)/(3)e^(-(t)/(3))dtF(x) = \int_{0}^{x} \frac{1}{3}e^{-\frac{t}{3}} \, dtF(x)=0x13et3dt
Using the integration technique, we get:
F ( x ) = [ 1 3 e t 3 ] 0 x = 1 3 e x 3 + 1 3 F ( x ) = 1 3 e t 3 0 x = 1 3 e x 3 + 1 3 F(x)=[-(1)/(3)e^(-(t)/(3))]_(0)^(x)=-(1)/(3)e^(-(x)/(3))+(1)/(3)F(x) = \left[-\frac{1}{3}e^{-\frac{t}{3}} \right]_{0}^{x} = -\frac{1}{3}e^{-\frac{x}{3}} + \frac{1}{3}F(x)=[13et3]0x=13ex3+13
So, the cumulative distribution function for x x xxx is:
F ( x ) = 1 e x 3 F ( x ) = 1 e x 3 F(x)=1-e^(-(x)/(3))F(x) = 1 - e^{-\frac{x}{3}}F(x)=1ex3
d) To find the probability that a student spends less than 4 hours studying for the exam, we can use the CDF calculated in part-c
P ( x < 4 ) = F ( 4 ) = 1 e 4 3 0.63 P ( x < 4 ) = F ( 4 ) = 1 e 4 3 0.63 P(x < 4)=F(4)=1-e^(-(4)/(3))~~0.63P(x < 4) = F(4) = 1 - e^{-\frac{4}{3}} \approx 0.63P(x<4)=F(4)=1e430.63
Thus, the probability that a student spends less than 4 hours studying for the exam is approximately 0.63.
e) To calculate the probability that a student spends more than 6 hours studying for the exam, we can use the complement rule:
P ( x > 6 ) = 1 P ( x 6 ) = 1 F ( 6 ) = 1 ( 1 e 6 3 ) = e 2 0.135 P ( x > 6 ) = 1 P ( x 6 ) = 1 F ( 6 ) = 1 ( 1 e 6 3 ) = e 2 0.135 P(x > 6)=1-P(x <= 6)=1-F(6)=1-(1-e^(-(6)/(3)))=e^(-2)~~0.135P(x > 6) = 1 - P(x \leq 6) = 1 - F(6) = 1 - (1 - e^{-\frac{6}{3}}) = e^{-2} \approx 0.135P(x>6)=1P(x6)=1F(6)=1(1e63)=e20.135
So, the probability that a student spends more than 6 hours studying for the exam is approximately 0.135.
\,
\,

Expectation (E) of a Continuous Random Variable

In probability theory, the expectation of a continuous random variable is a fundamental concept that represents the average value or mean of the variable under its probability distribution. It is also commonly referred to as the expected value. The expectation is a measure of the central tendency of the random variable, providing valuable insights into its typical or most likely value.
For a continuous random variable X X XXX, the expectation E ( X ) E ( X ) E(X)E(X)E(X) is calculated by taking the weighted average of all possible values that X X XXX can take, where the weights are determined by the probability density function (PDF) of X X XXX. The concept of expectation allows us to summarize the behavior of the random variable in a single numerical value.
For a continuous random variable X X XXX with probability density function f ( x ) f ( x ) f(x)f(x)f(x), the expectation E ( X ) E ( X ) E(X)E(X)E(X) is calculated as the integral of X X XXX multiplied by its probability density function over the entire range of X X XXX:
E ( X ) = M e a n ( μ ) = [ X f ( x ) ] d x E ( X ) = M e a n ( μ ) = [ X f ( x ) ] d x E(X)=Mean(mu)=int[X*f(x)]dxE(X) = Mean(\mu)=\int [X \cdot f(x)] dxE(X)=Mean(μ)=[Xf(x)]dx
Here, E ( X ) E ( X ) E(X)E(X)E(X) represents the expectation or expected value of the continuous random variable X X XXX. The integral int\int is taken over the entire range of X X XXX, and the product of X X XXX and f ( x ) f ( x ) f(x)f(x)f(x) is integrated with respect to the variable x x xxx.
Expectation of X 2 X 2 X^(2)X^2X2 (E( X 2 X 2 X^(2)X^2X2)) for a Continuous Random Variable:
For a continuous random variable X X XXX with probability density function f ( x ) f ( x ) f(x)f(x)f(x), the expectation of X 2 X 2 X^(2)X^2X2 (E( X 2 X 2 X^(2)X^2X2)) is calculated as the integral of X 2 X 2 X^(2)X^2X2 multiplied by its probability density function over the entire range of X X XXX:
E ( X 2 ) = [ X 2 f ( x ) ] d x E ( X 2 ) = [ X 2 f ( x ) ] d x E(X^(2))=int[X^(2)*f(x)]dxE(X^2) = \int [X^2 \cdot f(x)] dxE(X2)=[X2f(x)]dx
Variance (Var) for a Continuous Random Variable:
The variance (Var) of a continuous random variable X X XXX measures the spread or dispersion of X X XXX around its expectation E ( X ) E ( X ) E(X)E(X)E(X). It is calculated as follows:
V a r ( X ) = E ( X 2 ) [ E ( X ) ] 2 V a r ( X ) = E ( X 2 ) [ E ( X ) ] 2 Var(X)=E(X^(2))-[E(X)]^(2)Var(X) = E(X^2) - [E(X)]^2Var(X)=E(X2)[E(X)]2
Expectation of a Function of a Random Variable ( E ( f ( X ) ) E ( f ( X ) ) E(f(X))E(f(X))E(f(X))) for a Continuous Random Variable:
For a function f ( X ) f ( X ) f(X)f(X)f(X) of a continuous random variable X X XXX with probability density function f ( x ) f ( x ) f(x)f(x)f(x), the expectation of f ( X ) f ( X ) f(X)f(X)f(X) ( E ( f ( X ) ) E ( f ( X ) ) E(f(X))E(f(X))E(f(X))) is calculated as follows:
E ( f ( X ) ) = [ f ( x ) f ( x ) ] d x E ( f ( X ) ) = [ f ( x ) f ( x ) ] d x E(f(X))=int[f(x)*f(x)]dxE(f(X)) = \int [f(x) \cdot f(x)] dxE(f(X))=[f(x)f(x)]dx
\,
\,

Moment Generating Function (MGF) for Continuous Distribution

The Moment Generating Function (MGF) for a continuous random variable is a mathematical function that uniquely characterizes the random variable and provides a way to find its moments. It is defined as the expected value of the exponential function raised to the power of the random variable multiplied by a constant t t ttt. For a continuous random variable X X XXX with probability density function (PDF) f ( x ) f ( x ) f(x)f(x)f(x), the MGF is given by:
M X ( t ) = E [ e t X ] = e t x f ( x ) d x M X ( t ) = E [ e t X ] = e t x f ( x ) d x M_(X)(t)=E[e^(tX)]=int_(-oo)^(oo)e^(tx)*f(x)dxM_X(t) = E[e^{tX}] = \int_{-\infty}^{\infty} e^{tx} \cdot f(x) \, dxMX(t)=E[etX]=etxf(x)dx
where the integration is taken over the entire range of the random variable X X XXX.
The MGF has several important properties:
  1. The MGF is defined for all values of t t ttt in an interval containing zero.
  2. The n n nnnth moment of the random variable can be obtained by differentiating the MGF n n nnn times with respect to t t ttt and setting t = 0 t = 0 t=0t = 0t=0. Specifically, the n n nnnth moment is given by E [ X n ] = M X ( n ) ( 0 ) E [ X n ] = M X ( n ) ( 0 ) E[X^(n)]=M_(X)^((n))(0)E[X^n] = M_X^{(n)}(0)E[Xn]=MX(n)(0), where M X ( n ) ( 0 ) M X ( n ) ( 0 ) M_(X)^((n))(0)M_X^{(n)}(0)MX(n)(0) represents the n n nnnth derivative of M X ( t ) M X ( t ) M_(X)(t)M_X(t)MX(t) evaluated at t = 0 t = 0 t=0t = 0t=0.
  3. If the MGF exists and is finite in a neighborhood of zero, then it uniquely determines the probability distribution of the random variable X X XXX.
The MGF is particularly useful in finding moments of a random variable, especially in cases where finding the moments directly from the PDF may be difficult or cumbersome. The MGF can be applied to various types of continuous distributions, including the normal distribution, exponential distribution, gamma distribution, and many others.
However, it's important to note that not all continuous random variables have a valid MGF for all values of t t ttt, especially when the distribution lacks certain moments or the MGF does not converge in certain regions.
Statistic Formula
Mean E ( X ) = [ X f ( x ) ] d x E ( X ) = [ X f ( x ) ] d x E(X)=int[X*f(x)]dxE(X) = \int [X \cdot f(x)] dxE(X)=[Xf(x)]dx
Median (For symmetric distributions, the median is the same as the mean)
Mode (The mode is the value of x x xxx that maximizes the probability density function f ( x ) f ( x ) f(x)f(x)f(x))
Variance V a r ( X ) = E ( X 2 ) [ E ( X ) ] 2 V a r ( X ) = E ( X 2 ) [ E ( X ) ] 2 Var(X)=E(X^(2))-[E(X)]^(2)Var(X) = E(X^2) - [E(X)]^2Var(X)=E(X2)[E(X)]2
Standard Deviation S D ( X ) = V a r ( X ) S D ( X ) = V a r ( X ) SD(X)=sqrt(Var(X))SD(X) = \sqrt{Var(X)}SD(X)=Var(X)
Statistic Formula Mean E(X)=int[X*f(x)]dx Median (For symmetric distributions, the median is the same as the mean) Mode (The mode is the value of x that maximizes the probability density function f(x)) Variance Var(X)=E(X^(2))-[E(X)]^(2) Standard Deviation SD(X)=sqrt(Var(X))| **Statistic** | **Formula** | | :---: | :---: | | Mean | $E(X) = \int [X \cdot f(x)] dx$ | | Median | (For symmetric distributions, the median is the same as the mean) | | Mode | (The mode is the value of $x$ that maximizes the probability density function $f(x)$) | | Variance | $Var(X) = E(X^2) - [E(X)]^2$ | | Standard Deviation | $SD(X) = \sqrt{Var(X)}$ |
=>\Rightarrow This allows you to investigate Probability Density Functions (PDFs)
Enter a PDF in the f(x) input box and the range it operates on underneath.
(note this applet works only for single finite ranges).
\,

Linearity of Expectation

The linearity of expectation is a fundamental property that holds for both discrete and continuous random variables. It states that the expectation (or mean) of a linear combination of random variables is equal to the linear combination of their individual expectations.
For any constants a a aaa and b b bbb and any random variables X X XXX and Y Y YYY, the linearity of expectation can be expressed as follows:
  1. For a single random variable X X XXX:
E [ a X + b ] = a E [ X ] + b E [ a X + b ] = a E [ X ] + b E[aX+b]=aE[X]+bE[aX + b] = aE[X] + bE[aX+b]=aE[X]+b
  1. For multiple random variables X X XXX and Y Y YYY:
E [ a X + b Y ] = a E [ X ] + b E [ Y ] E [ a X + b Y ] = a E [ X ] + b E [ Y ] E[aX+bY]=aE[X]+bE[Y]E[aX + bY] = aE[X] + bE[Y]E[aX+bY]=aE[X]+bE[Y]
The linearity of expectation is a powerful property that simplifies the calculation of the expected value when dealing with combinations of random variables. It allows us to break down complex expressions into simpler components and compute their expectations separately.
Question:
A continuous random variable X X XXX is described by the probability density function (PDF) given by:
f ( x ) = { 2 9 x 0 x 3 0 otherwise f ( x ) = 2 9 x 0 x 3 0 otherwise f(x)={[(2)/(9)x,0 <= x <= 3],[0,"otherwise"]:}f(x) = \begin{cases} \frac{2}{9}x & 0 \leq x \leq 3 \\ 0 & \text{otherwise} \end{cases}f(x)={29x0x30otherwise
Calculate the following:
a) The expectation ( E [ X ] E [ X ] E[X]E[X]E[X]) of the random variable X X XXX.
b) The expectation ( E [ X 2 ] E [ X 2 ] E[X^(2)]E[X^2]E[X2]) of the random variable X 2 X 2 X^(2)X^2X2.
c) The variance ( V a r ( X ) V a r ( X ) Var(X)Var(X)Var(X)) of the random variable X X XXX.
d) The expectation ( E [ f ( X ) ] E [ f ( X ) ] E[f(X)]E[f(X)]E[f(X)]) of the function f ( X ) = 2 X + 1 f ( X ) = 2 X + 1 f(X)=2X+1f(X) = 2X + 1f(X)=2X+1.
 Solution:
a) The expectation ( E [ X ] E [ X ] E[X]E[X]E[X]) of a continuous random variable X X XXX is calculated by integrating the product of the variable X X XXX and its probability density function (PDF) over the entire range of possible values.
For the given probability density function f ( x ) f ( x ) f(x)f(x)f(x), we have:
E [ X ] = x f ( x ) d x E [ X ] = x f ( x ) d x E[X]=int_(-oo)^(oo)x*f(x)dxE[X] = \int_{-\infty}^{\infty} x \cdot f(x) \, dxE[X]=xf(x)dx
Since f ( x ) = 0 f ( x ) = 0 f(x)=0f(x) = 0f(x)=0 for x < 0 x < 0 x < 0x < 0x<0 and x > 3 x > 3 x > 3x > 3x>3, we can limit the integration to the range where f ( x ) f ( x ) f(x)f(x)f(x) is nonzero, which is from 0 to 3.
E [ X ] = 0 3 x ( 2 9 x ) d x E [ X ] = 0 3 x 2 9 x d x E[X]=int_(0)^(3)x*((2)/(9)x)dxE[X] = \int_{0}^{3} x \cdot \left(\frac{2}{9}x\right) \, dxE[X]=03x(29x)dx
Now, we can perform the integration:
E [ X ] = 2 9 0 3 x 2 d x E [ X ] = 2 9 0 3 x 2 d x E[X]=(2)/(9)int_(0)^(3)x^(2)dxE[X] = \frac{2}{9} \int_{0}^{3} x^2 \, dxE[X]=2903x2dx
E [ X ] = 2 9 [ x 3 3 ] 0 3 E [ X ] = 2 9 x 3 3 0 3 E[X]=(2)/(9)[(x^(3))/(3)]_(0)^(3)E[X] = \frac{2}{9} \left[\frac{x^3}{3}\right]_{0}^{3}E[X]=29[x33]03
E [ X ] = 2 9 ( 3 3 3 0 3 3 ) E [ X ] = 2 9 3 3 3 0 3 3 E[X]=(2)/(9)((3^(3))/(3)-(0^(3))/(3))E[X] = \frac{2}{9} \left(\frac{3^3}{3} - \frac{0^3}{3}\right)E[X]=29(333033)
E [ X ] = 2 9 9 E [ X ] = 2 9 9 E[X]=(2)/(9)*9E[X] = \frac{2}{9} \cdot 9E[X]=299
E [ X ] = 2 E [ X ] = 2 E[X]=2E[X] = 2E[X]=2
Therefore, the expectation ( E [ X ] E [ X ] E[X]E[X]E[X]) of the random variable X X XXX is 2.
b) To find the expectation ( E [ X 2 ] E [ X 2 ] E[X^(2)]E[X^2]E[X2]), we follow a similar approach:
E [ X 2 ] = x 2 f ( x ) d x E [ X 2 ] = x 2 f ( x ) d x E[X^(2)]=int_(-oo)^(oo)x^(2)*f(x)dxE[X^2] = \int_{-\infty}^{\infty} x^2 \cdot f(x) \, dxE[X2]=x2f(x)dx
Since f ( x ) = 0 f ( x ) = 0 f(x)=0f(x) = 0f(x)=0 for x < 0 x < 0 x < 0x < 0x<0 and x > 3 x > 3 x > 3x > 3x>3, we integrate over the range from 0 to 3:
E [ X 2 ] = 0 3 x 2 ( 2 9 x ) d x E [ X 2 ] = 0 3 x 2 2 9 x d x E[X^(2)]=int_(0)^(3)x^(2)*((2)/(9)x)dxE[X^2] = \int_{0}^{3} x^2 \cdot \left(\frac{2}{9}x\right) \, dxE[X2]=03x2(29x)dx
E [ X 2 ] = 2 9 0 3 x 3 d x E [ X 2 ] = 2 9 0 3 x 3 d x E[X^(2)]=(2)/(9)int_(0)^(3)x^(3)dxE[X^2] = \frac{2}{9} \int_{0}^{3} x^3 \, dxE[X2]=2903x3dx
E [ X 2 ] = 2 9 [ x 4 4 ] 0 3 E [ X 2 ] = 2 9 x 4 4 0 3 E[X^(2)]=(2)/(9)[(x^(4))/(4)]_(0)^(3)E[X^2] = \frac{2}{9} \left[\frac{x^4}{4}\right]_{0}^{3}E[X2]=29[x44]03
E [ X 2 ] = 2 9 ( 3 4 4 0 4 4 ) E [ X 2 ] = 2 9 3 4 4 0 4 4 E[X^(2)]=(2)/(9)((3^(4))/(4)-(0^(4))/(4))E[X^2] = \frac{2}{9} \left(\frac{3^4}{4} - \frac{0^4}{4}\right)E[X2]=29(344044)
E [ X 2 ] = 2 9 81 4 E [ X 2 ] = 2 9 81 4 E[X^(2)]=(2)/(9)*(81)/(4)E[X^2] = \frac{2}{9} \cdot \frac{81}{4}E[X2]=29814
E [ X 2 ] = 9 2 E [ X 2 ] = 9 2 E[X^(2)]=(9)/(2)E[X^2] = \frac{9}{2}E[X2]=92
Therefore, the expectation ( E [ X 2 ] E [ X 2 ] E[X^(2)]E[X^2]E[X2]) of the random variable X 2 X 2 X^(2)X^2X2 is 9 2 9 2 (9)/(2)\frac{9}{2}92.
c) The variance ( V a r ( X ) V a r ( X ) Var(X)Var(X)Var(X)) of a random variable X X XXX is defined as:
V a r ( X ) = E [ X 2 ] ( E [ X ] ) 2 V a r ( X ) = E [ X 2 ] ( E [ X ] ) 2 Var(X)=E[X^(2)]-(E[X])^(2)Var(X) = E[X^2] - (E[X])^2Var(X)=E[X2](E[X])2
From parts (a) and (b), we have:
V a r ( X ) = 9 2 2 2 V a r ( X ) = 9 2 2 2 Var(X)=(9)/(2)-2^(2)Var(X) = \frac{9}{2} - 2^2Var(X)=9222
V a r ( X ) = 9 2 4 V a r ( X ) = 9 2 4 Var(X)=(9)/(2)-4Var(X) = \frac{9}{2} - 4Var(X)=924
V a r ( X ) = 1 2 V a r ( X ) = 1 2 Var(X)=(1)/(2)Var(X) = \frac{1}{2}Var(X)=12
Therefore, the variance ( V a r ( X ) V a r ( X ) Var(X)Var(X)Var(X)) of the random variable X X XXX is 1 2 1 2 (1)/(2)\frac{1}{2}12.
d) To find the expectation ( E [ f ( X ) ] E [ f ( X ) ] E[f(X)]E[f(X)]E[f(X)]) of the function f ( X ) = 2 X + 1 f ( X ) = 2 X + 1 f(X)=2X+1f(X) = 2X + 1f(X)=2X+1, we apply the transformation to the random variable X X XXX and calculate its expectation:
E [ f ( X ) ] = E [ 2 X + 1 ] E [ f ( X ) ] = E [ 2 X + 1 ] E[f(X)]=E[2X+1]E[f(X)] = E[2X + 1]E[f(X)]=E[2X+1]
Using linearity of expectation:
E [ f ( X ) ] = 2 E [ X ] + E [ 1 ] E [ f ( X ) ] = 2 E [ X ] + E [ 1 ] E[f(X)]=2E[X]+E[1]E[f(X)] = 2E[X] + E[1]E[f(X)]=2E[X]+E[1]
From part (a),
E [ f ( X ) ] = 2 × 2 + 1 E [ f ( X ) ] = 2 × 2 + 1 E[f(X)]=2xx2+1E[f(X)] = 2\times2 + 1E[f(X)]=2×2+1
E [ f ( X ) ] = 5 E [ f ( X ) ] = 5 E[f(X)]=5E[f(X)] = 5E[f(X)]=5
\,
\,

Normal Distribution X N ( μ , σ 2 ) X N ( μ , σ 2 ) X∼N(mu,sigma^(2))X\sim N(\mu,\sigma ^{2})XN(μ,σ2)

The normal distribution, also known as the Gaussian distribution, is one of the most widely used probability distributions in statistics. It is a continuous distribution characterized by a symmetric bell-shaped curve. The shape of the curve is determined by two parameters: the mean ( μ μ mu\muμ) and the standard deviation ( σ σ sigma\sigmaσ).
The probability density function (PDF) of the normal distribution is given by:
f ( x ) = 1 σ 2 π exp ( ( x μ ) 2 2 σ 2 ) f ( x ) = 1 σ 2 π exp ( x μ ) 2 2 σ 2 f(x)=(1)/((sigmasqrt(2pi)))exp(-((x-mu)^(2))/(2sigma^(2)))f(x) = \frac{1}{{\sigma \sqrt{2\pi}}} \exp\left(-\frac{(x-\mu)^2}{2\sigma^2}\right)f(x)=1σ2πexp((xμ)22σ2)
where:
  • x x xxx is the random variable.
  • μ μ mu\muμ is the mean or expected value, which represents the center of the distribution.
  • σ σ sigma\sigmaσ is the standard deviation, which measures the spread or dispersion of the data.
  • π π pi\piπ is a mathematical constant (approximately 3.14159).
The normal distribution is characterized by the following properties:
  1. It is symmetric around the mean μ μ mu\muμ.
  2. The curve reaches its maximum at x = μ x = μ x=mux = \mux=μ.
  3. The total area under the curve is equal to 1, representing the probabilities of all possible outcomes.
Applications of the normal distribution include modeling of measurement errors, natural phenomena, financial data, and many other real-world processes. Its widespread use is due to its mathematical tractability and its ability to model a wide range of data with just two parameters.
The standard normal distribution, denoted by Z N ( 0 , 1 ) Z N ( 0 , 1 ) Z∼N(0,1)Z \sim \mathcal{N}(0, 1)ZN(0,1), is a special case of the normal distribution with a mean μ = 0 μ = 0 mu=0\mu=0μ=0 and a standard deviation σ = 1 σ = 1 sigma=1\sigma=1σ=1. It has the probability density function:
ϕ ( z ) = 1 2 π e z 2 2 ϕ ( z ) = 1 2 π e z 2 2 phi(z)=(1)/(sqrt(2pi))*e^(-(z^(2))/(2))\phi(z) = \frac{1}{\sqrt{2\pi}} \cdot e^{-\frac{z^2}{2}}ϕ(z)=12πez22
To convert a value x x xxx from a normal distribution ( X X XXX) to a standard normal distribution ( Z Z ZZZ), we use the transformation called standardization or normalization. The formula for this transformation is known as the z-score formula:
Z = X μ σ Z = X μ σ Z=(X-mu)/(sigma)Z = \frac{X - \mu}{\sigma}Z=Xμσ
where:
  • X X XXX is the value from the original normal distribution,
  • μ μ mu\muμ is the mean of the original normal distribution,
  • σ σ sigma\sigmaσ is the standard deviation of the original normal distribution, and
  • Z Z ZZZ is the transformed value representing the z-score in the standard normal distribution.
The z-score, Z Z ZZZ, represents the number of standard deviations the original value X X XXX is from the mean μ μ mu\muμ of the original distribution. If Z Z ZZZ is positive, the X X XXX value is above the mean; if it is negative, the X X XXX value is below the mean.
The standard normal distribution is particularly useful because it simplifies various calculations and allows us to use standard normal distribution tables or statistical software to find probabilities and percentiles for specific z-scores. These tables and software provide the cumulative probabilities, allowing us to determine the probability of obtaining values within a certain range or greater than a specific value in a normal distribution without explicitly calculating integrals.
In summary, converting a normal distribution to a standard normal distribution involves standardizing data using the z-score formula. This transformation is valuable in statistical analysis, hypothesis testing, and making comparisons across different normal distributions, as it allows us to work with a common standardized scale.
Question:
The heights of adult male students in a university are normally distributed with a mean of 175 cm and a standard deviation of 7 cm.
a) What percentage of male students have heights between 165 cm and 185 cm?
b) If the university sets a height requirement of being in the top 10% of male students, what is the minimum height a male student must have to meet this requirement?
 Solution:
a) To find the percentage of male students with heights between 165 cm and 185 cm, we will use the properties of the normal distribution. Let's first standardize the values using the z-score formula:
z = x μ σ z = x μ σ z=(x-mu)/(sigma)z = \frac{x - \mu}{\sigma}z=xμσ
where:
  • x x xxx is the height value,
  • μ μ mu\muμ is the mean height (175 cm),
  • σ σ sigma\sigmaσ is the standard deviation (7 cm), and
  • z z zzz is the z-score.
For x = 165 x = 165 x=165x = 165x=165 cm:
z 1 = 165 175 7 = 1.43 z 1 = 165 175 7 = 1.43 z_(1)=(165-175)/(7)=-1.43z_1 = \frac{165 - 175}{7} = -1.43z1=1651757=1.43
For x = 185 x = 185 x=185x = 185x=185 cm:
z 2 = 185 175 7 = 1.43 z 2 = 185 175 7 = 1.43 z_(2)=(185-175)/(7)=1.43z_2 = \frac{185 - 175}{7} = 1.43z2=1851757=1.43
Now, we need to find the area under the standard normal curve between z 1 z 1 z_(1)z_1z1 and z 2 z 2 z_(2)z_2z2. We can use a standard normal distribution table or a calculator to find these values. Since the normal distribution is symmetric, the area between z 1 z 1 z_(1)z_1z1 and z 2 z 2 z_(2)z_2z2 is the same as the area between z 2 z 2 -z_(2)-z_2z2 and z 1 z 1 -z_(1)-z_1z1. Therefore, we find the area for z = 1.43 z = 1.43 z=1.43z = 1.43z=1.43 and then subtract it from 1:
P ( 1.43 < z < 1.43 ) = 1 2 × P ( z > 1.43 ) P ( 1.43 < z < 1.43 ) = 1 2 × P ( z > 1.43 ) P(-1.43 < z < 1.43)=1-2xx P(z > 1.43)P(-1.43 < z < 1.43) = 1 - 2 \times P(z > 1.43)P(1.43<z<1.43)=12×P(z>1.43)
Using a standard normal distribution table or calculator, we find P ( z > 1.43 ) 0.0764 P ( z > 1.43 ) 0.0764 P(z > 1.43)~~0.0764P(z > 1.43) \approx 0.0764P(z>1.43)0.0764.
P ( 1.43 < z < 1.43 ) 1 2 × 0.0764 0.8472 P ( 1.43 < z < 1.43 ) 1 2 × 0.0764 0.8472 P(-1.43 < z < 1.43)~~1-2xx0.0764~~0.8472P(-1.43 < z < 1.43) \approx 1 - 2 \times 0.0764 \approx 0.8472P(1.43<z<1.43)12×0.07640.8472
So, approximately 84.72% of male students have heights between 165 cm and 185 cm.
b) To find the minimum height a male student must have to be in the top 10% of male students, we need to find the z-score corresponding to the 90th percentile. In other words, we need to find the z-score such that P ( z < z score ) = 0.90 P ( z < z score ) = 0.90 P(z < z_("score"))=0.90P(z < z_{\text{score}}) = 0.90P(z<zscore)=0.90.
Using a standard normal distribution table or calculator, we find the z-score corresponding to the 90th percentile is approximately 1.28.
Now, we can use the z-score formula to find the height corresponding to this z-score:
z score = x 175 7 z score = x 175 7 z_("score")=(x-175)/(7)z_{\text{score}} = \frac{x - 175}{7}zscore=x1757
1.28 = x 175 7 1.28 = x 175 7 1.28=(x-175)/(7)1.28 = \frac{x - 175}{7}1.28=x1757
Solving for x x xxx:
x 175 = 1.28 × 7 x 175 = 1.28 × 7 x-175=1.28 xx7x - 175 = 1.28 \times 7x175=1.28×7
x 184 x 184 x~~184x \approx 184x184
So, the minimum height a male student must have to meet the university's requirement of being in the top 10% of male students is approximately 184 cm.
\,
Question:
The amount of time it takes a factory worker to assemble a particular product is normally distributed with a mean of 20 minutes and a standard deviation of 3 minutes.
a) What percentage of the workers take more than 25 minutes to assemble the product?
b) If the factory wants to identify the fastest 20% of workers, what is the maximum time a worker can take to be in this category?
 Solution:
a) To find the percentage of workers who take more than 25 minutes to assemble the product, we use the properties of the normal distribution. Let's first standardize the value using the z-score formula:
z = x μ σ z = x μ σ z=(x-mu)/(sigma)z = \frac{x - \mu}{\sigma}z=xμσ
where:
  • x x xxx is the time taken to assemble the product (in minutes),
  • μ μ mu\muμ is the mean time (20 minutes),
  • σ σ sigma\sigmaσ is the standard deviation (3 minutes), and
  • z z zzz is the z-score.
For x = 25 x = 25 x=25x = 25x=25 minutes:
z = 25 20 3 = 1.67 z = 25 20 3 = 1.67 z=(25-20)/(3)=1.67z = \frac{25 - 20}{3} = 1.67z=25203=1.67
Now, we need to find the area under the standard normal curve to the right of this z-score (since we want values greater than 25 minutes). We can use a standard normal distribution table or a calculator to find this area.
P ( z > 1.67 ) P ( z > 1.67 ) P(z > 1.67)P(z > 1.67)P(z>1.67)
Using a standard normal distribution table or calculator, we find P ( z > 1.67 ) 0.0475 P ( z > 1.67 ) 0.0475 P(z > 1.67)~~0.0475P(z > 1.67) \approx 0.0475P(z>1.67)0.0475.
So, approximately 4.75% of workers take more than 25 minutes to assemble the product.
b) To find the maximum time a worker can take to be in the fastest 20% of workers, we need to find the z-score corresponding to the 80th percentile. In other words, we need to find the z-score such that P ( z < z score ) = 0.80 P ( z < z score ) = 0.80 P(z < z_("score"))=0.80P(z < z_{\text{score}}) = 0.80P(z<zscore)=0.80.
Using a standard normal distribution table or calculator, we find the z-score corresponding to the 80th percentile is approximately 0.84.
Now, we can use the z-score formula to find the time corresponding to this z-score:
z score = x 20 3 z score = x 20 3 z_("score")=(x-20)/(3)z_{\text{score}} = \frac{x - 20}{3}zscore=x203
0.84 = x 20 3 0.84 = x 20 3 0.84=(x-20)/(3)0.84 = \frac{x - 20}{3}0.84=x203
Solving for x x xxx:
x 20 = 0.84 × 3 x 20 = 0.84 × 3 x-20=0.84 xx3x - 20 = 0.84 \times 3x20=0.84×3
x 22.52 x 22.52 x~~22.52x \approx 22.52x22.52
So, the maximum time a worker can take to be in the fastest 20% of workers is approximately 22.52 minutes.

Gamma Distribution

The gamma distribution is a continuous probability distribution that is commonly used to model positive-valued random variables. It is a flexible distribution that can represent a wide range of shapes, making it useful in various statistical applications.
The probability density function (PDF) of the gamma distribution is given by:
f ( x ) = 1 Γ ( k ) θ k x k 1 e x θ f ( x ) = 1 Γ ( k ) θ k x k 1 e x θ f(x)=(1)/(Gamma(k)*theta ^(k))*x^(k-1)*e^(-(x)/( theta))f(x) = \frac{1}{\Gamma(k) \cdot \theta^k} \cdot x^{k-1} \cdot e^{-\frac{x}{\theta}}f(x)=1Γ(k)θkxk1exθ
where:
  • x 0 x 0 x >= 0x \geq 0x0 is the random variable.
  • k > 0 k > 0 k > 0k > 0k>0 is the shape parameter.
  • θ > 0 θ > 0 theta > 0\theta > 0θ>0 is the scale parameter.
  • Γ ( k ) Γ ( k ) Gamma(k)\Gamma(k)Γ(k) is the gamma function, which is a generalization of the factorial for non-integer values.
The gamma distribution includes several special cases, such as the exponential distribution (when k = 1 k = 1 k=1k = 1k=1), the chi-squared distribution (when k k kkk is an integer), and the Erlang distribution (when k k kkk is an integer and represents the number of events in an interval).
The mean and variance of the gamma distribution are k θ k θ k*thetak \cdot \thetakθ and k θ 2 k θ 2 k*theta^(2)k \cdot \theta^2kθ2, respectively.
The gamma distribution is often used to model the time between events in queuing systems, the lifetime of certain materials, and the waiting time for arrival processes, among other applications.
To fit the gamma distribution to data or estimate its parameters, various statistical methods such as maximum likelihood estimation (MLE) and method of moments can be used. The gamma distribution is widely supported in statistical software packages and provides a valuable tool for modeling positive-valued random variables in various fields of study.

Exponential Distribution

The exponential distribution is a continuous probability distribution commonly used to model the time between events in a Poisson process. It is often employed to describe the waiting times between successive events when events occur at a constant rate independently of the time since the last event.
The probability density function (PDF) of the exponential distribution is given by:
f ( x ) = λ e λ x f ( x ) = λ e λ x f(x)=lambdae^(-lambda x)f(x) = \lambda e^{-\lambda x}f(x)=λeλx
where:
  • x 0 x 0 x >= 0x \geq 0x0 is the random variable representing the waiting time between events.
  • λ > 0 λ > 0 lambda > 0\lambda > 0λ>0 is the rate parameter, which represents the average number of events occurring per unit of time.
The cumulative distribution function (CDF) of the exponential distribution is:
F ( x ) = 1 e λ x F ( x ) = 1 e λ x F(x)=1-e^(-lambda x)F(x) = 1 - e^{-\lambda x}F(x)=1eλx
The mean (expected value) and variance of the exponential distribution are both 1 λ 1 λ (1)/(lambda)\frac{1}{\lambda}1λ.
The exponential distribution has the memoryless property, meaning that the probability of an event occurring in the next time interval is independent of how much time has already passed. This property is expressed as:
P ( X > s + t | X > s ) = P ( X > t ) P ( X > s + t | X > s ) = P ( X > t ) P(X > s+t|X > s)=P(X > t)P(X > s + t | X > s) = P(X > t)P(X>s+t|X>s)=P(X>t)
where X X XXX is a random variable representing the waiting time between events, and s s sss and t t ttt are non-negative time values.
The exponential distribution is used in various fields, including queuing theory, reliability analysis, telecommunications, and survival analysis, among others. It provides a simple and versatile model for situations involving the time between events occurring at a constant rate.
Exponentiated Exponential Distribution (not in syllabus):
The exponentiated exponential distribution is a continuous probability distribution that extends the exponential distribution by introducing an additional parameter. It is also known as the generalized exponential distribution or the double exponential distribution. The exponentiated exponential distribution has two parameters, a a aaa and λ λ lambda\lambdaλ, representing the shape and rate parameters, respectively.
The probability density function (PDF) of the exponentiated exponential distribution is given by:
f ( x ) = a λ 2 exp ( a λ | x | ) f ( x ) = a λ 2 exp ( a λ | x | ) f(x)=(a lambda)/(2)exp(-a lambda|x|)f(x) = \frac{a \lambda}{2} \exp(-a \lambda |x|)f(x)=aλ2exp(aλ|x|)
where x x xxx can be any real value.
The exponentiated exponential distribution allows for more flexibility in modeling data compared to the standard exponential distribution. The value of the parameter a a aaa determines the shape of the distribution. When a = 1 a = 1 a=1a = 1a=1, the exponentiated exponential distribution reduces to the standard exponential distribution.